What is Decidable about Partially Observable Markov Decision Processes with ω-Regular Objectives
نویسندگان
چکیده
We consider partially observable Markov decision processes (POMDPs) with ω-regular conditions specified as parity objectives. The class of ω-regular languages extends regular languages to infinite strings and provides a robust specification language to express all properties used in verification, and parity objectives are canonical forms to express ω-regular conditions. The qualitative analysis problem given a POMDP and a parity objective asks whether there is a strategy to ensure that the objective is satisfied with probability 1 (resp. positive probability). While the qualitative analysis problems are known to be undecidable even for very special cases of parity objectives, we establish decidability (with optimal complexity) of the qualitative analysis problems for POMDPs with all parity objectives under finitememory strategies. We establish asymptotically optimal (exponential) memory bounds and EXPTIMEcompleteness of the qualitative analysis problems under finite-memory strategies for POMDPs with parity objectives.
منابع مشابه
What is Decidable about Partially Observable Markov Decision Processes with omega-Regular Objectives
We consider partially observable Markov decision processes (POMDPs) with ω-regular conditions specified as parity objectives. The class of ω-regular languages extends regular languages to infinite strings and provides a robust specification language to express all properties used in verification, and parity objectives are canonical forms to express ω-regular conditions. The qualitative analysis...
متن کاملMultiple-Environment Markov Decision Processes
We introduce Multi-Environment Markov Decision Processes (MEMDPs) which are MDPs with a set of probabilistic transition functions. The goal in a MEMDP is to synthesize a single controller with guaranteed performances against all environments even though the environment is unknown a priori. While MEMDPs can be seen as a special class of partially observable MDPs, we show that several verificatio...
متن کاملDeciding the Value 1 Problem for ]-acyclic Partially Observable Markov Decision Processes
The value 1 problem is a natural decision problem in algorithmic game theory. For partially observable Markov decision processes with reachability objective, this problem is defined as follows: are there strategies that achieve the reachability objective with probability arbitrarily close to 1? This problem was shown undecidable recently. Our contribution is to introduce a class of partially ob...
متن کاملDeciding the Value 1 Problem for $\sharp$ -acyclic Partially Observable Markov Decision Processes
The value 1 problem is a natural decision problem in algorithmic game theory. For partially observable Markov decision processes with reachability objective, this problem is defined as follows: are there observational strategies that achieve the reachability objective with probability arbitrarily close to 1? This problem was shown undecidable recently. Our contribution is to introduce a class o...
متن کاملA Specification Logic for Reasoning about Actions and Explicit Observations
We propose a logic that can express explicit observation constants, as a bridge towards reasoning about action using partially observable Markov decision processes (POMDPs). We introduce a new modal operator [o | α]—informally, [o | α]φ means ‘after perceiving observation o, given α was performed, necessarily φ’. In this paper, we focus on the challenges concerning sensing with explicit observa...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- J. Comput. Syst. Sci.
دوره 82 شماره
صفحات -
تاریخ انتشار 2016